Speech and Gesture Multimodal Control of a Whole Earth 3D Visualization Environment
نویسندگان
چکیده
A growing body of research shows several advantages to multimodal interfaces including increased expressiveness, flexibility, and user freedom. This paper investigates the design of such an interface that integrates speech and hand gestures. The interface has the additional property of operating relative to the user and can be used while the user is in motion or standing at a distance from the computer display. The paper then describes an implementation of the multimodal interface for a whole Earth 3D visualization which presents navigation interface challenges due to the large magnitude of scale and extended spaces that are available. The characteristics of the multimodal interface are examined, such as speed, recognizability of gestures, ease and accuracy of use, and learnability under likely conditions of use. This implementation shows that such a multimodal interface can be effective in a real environment and sets some parameters for the design and use of such interfaces.
منابع مشابه
A Multi Server Multi User Approach for Distributed Virtual Environments
In this paper a VR architecture is described that enables an easy and very scalable way to build up multi user and multi wall projection systems for distributed applications and generic VR interface integration. The approach is based on a multi server architecture and a strict separation of two layers of control of the distributed virtual environment. The 3D representation is full VRML2.0 compl...
متن کاملA Multimodal Virtual Reality Interface for VTK
The object oriented Visualization Toolkit (VTK) is widely used for scientific visualization. VTK is a visualization library that provides functions for presenting 3D data. Interaction with the visualized data is done by mouse and keyboard. Support for three-dimensional and multimodal input is non-existent. This paper describes VR-VTK: a multimodal interface to VTK on a desktop virtual environme...
متن کاملCommunicating and Acting: Understanding Gesture in Simulation Semantics
In this paper, we introduce an architecture for multimodal communication between humans and computers engaged in a shared task. We describe a representative dialogue between an artificial agent and a human that will be demonstrated live during the presentation. This assumes a multimodal environment and semantics for facilitating communication and interaction with a computational agent. To this ...
متن کاملVIRSTORY: A Collaborative Virtual Storytelling
From the results of two Human behavior studies in small group interactions we constructed iteratively a Collaborative Virtual Environment named “VIRSTORY”. This system is a digital storytelling with speech and 3D gesture recognition technologies using like input devices. It includes several modules: multimodal interaction module, behavior module to animate autonomous expressive characters, etc.
متن کاملDimitra Anastasiou Survey on Speech , Machine Translation and Gestures in Ambient Assisted Living
In this paper we provide the state-of-the-art of existing proprietary and free and open source software (FOSS) automatic speech recognition (ASR), speech synthesizers, and Machine Translation (MT) tools. We also focus on the need for multimodal communication including gestures, furnishing some examples of 3D gesture recognition software. Our current experiment is based on interoperability betwe...
متن کامل